cosine distance
RGMDT: Return-Gap-MinimizingDecisionTree ExtractioninNon-EuclideanMetricSpace
In this paper, we establish an upper bound on the return gap between the oracle expert policy and an optimal decision tree policy. This enables us to recast the DT extraction problem into a novel non-euclidean clustering problem over the local observation and action values space of each agent, with action values as cluster labels and the upper bound on the return gap as clustering loss.
- North America > United States > Pennsylvania > Allegheny County > Pittsburgh (0.04)
- North America > United States > California > Monterey County > Monterey (0.04)
- Europe > Ireland > Leinster > County Dublin > Dublin (0.04)
- Europe > Finland > Northern Savo > Kuopio (0.04)
- Europe > Switzerland > Basel-City > Basel (0.04)
- Asia > Middle East > Israel (0.04)
- North America > United States > Maine > Cumberland County > Standish (0.14)
- North America > United States > California (0.05)
- Asia > India > Rajasthan (0.04)
- (9 more...)
- Health & Medicine (1.00)
- Education (0.93)
- Asia > China (0.04)
- North America > United States > New York > Broome County > Binghamton (0.04)
- North America > United States > Massachusetts > Hampshire County > Amherst (0.04)
- (3 more...)
Fubini Study geometry of representation drift in high dimensional data
High dimensional representation drift is commonly quantified using Euclidean or cosine distances, which presuppose fixed coordinates when comparing representations across time, training or preprocessing stages. While effective in many settings, these measures entangle intrinsic changes in the data with variations induced by arbitrary parametrizations. We introduce a projective geometric view of representation drift grounded in the Fubini Study metric, which identifies representations that differ only by gauge transformations such as global rescalings or sign flips. Applying this framework to empirical high dimensional datasets, we explicitly construct representation trajectories and track their evolution through cumulative geometric drift. Comparing Euclidean, cosine and Fubini Study distances along these trajectories reveals that conventional metrics systematically overestimate change whenever representations carry genuine projective ambiguity. By contrast, the Fubini Study metric isolates intrinsic evolution by remaining invariant under gauge-induced fluctuations. We further show that the difference between cosine and Fubini Study drift defines a computable, monotone quantity that directly captures representation churn attributable to gauge freedom. This separation provides a diagnostic for distinguishing meaningful structural evolution from parametrization artifacts, without introducing model-specific assumptions. Overall, we establish a geometric criterion for assessing representation stability in high-dimensional systems and clarify the limits of angular distances. Embedding representation dynamics in projective space connects data analysis with established geometric programs and yields observables that are directly testable in empirical workflows.
- North America > United States (0.28)
- Europe > Italy > Campania > Naples (0.04)
Convergence of Outputs When Two Large Language Models Interact in a Multi-Agentic Setup
Maiti, Aniruddha, Nimmagadda, Satya, Jammuladinne, Kartha Veerya, Sengupta, Niladri, Jana, Ananya
In this work, we report what happens when two large language models respond to each other for many turns without any outside input in a multi-agent setup. The setup begins with a short seed sentence. After that, each model reads the other's output and generates a response. This continues for a fixed number of steps. We used Mistral Nemo Base 2407 and Llama 2 13B hf. We observed that most conversations start coherently but later fall into repetition. In many runs, a short phrase appears and repeats across turns. Once repetition begins, both models tend to produce similar output rather than introducing a new direction in the conversation. This leads to a loop where the same or similar text is produced repeatedly. We describe this behavior as a form of convergence. It occurs even though the models are large, trained separately, and not given any prompt instructions. To study this behavior, we apply lexical and embedding-based metrics to measure how far the conversation drifts from the initial seed and how similar the outputs of the two models becomes as the conversation progresses.
- North America > United States > West Virginia > Cabell County > Huntington (0.04)
- Asia > India (0.04)